14 research outputs found

    A Deep-Learning Model for Subject-Independent Human Emotion Recognition Using Electrodermal Activity Sensors

    No full text
    One of the main objectives of Active and Assisted Living (AAL) environments is to ensure that elderly and/or disabled people perform/live well in their immediate environments; this can be monitored by among others the recognition of emotions based on non-highly intrusive sensors such as Electrodermal Activity (EDA) sensors. However, designing a learning system or building a machine-learning model to recognize human emotions while training the system on a specific group of persons and testing the system on a totally a new group of persons is still a serious challenge in the field, as it is possible that the second testing group of persons may have different emotion patterns. Accordingly, the purpose of this paper is to contribute to the field of human emotion recognition by proposing a Convolutional Neural Network (CNN) architecture which ensures promising robustness-related results for both subject-dependent and subject-independent human emotion recognition. The CNN model has been trained using a grid search technique which is a model hyperparameter optimization technique to fine-tune the parameters of the proposed CNN architecture. The overall concept’s performance is validated and stress-tested by using MAHNOB and DEAP datasets. The results demonstrate a promising robustness improvement regarding various evaluation metrics. We could increase the accuracy for subject-independent classification to 78% and 82% for MAHNOB and DEAP respectively and to 81% and 85% subject-dependent classification for MAHNOB and DEAP respectively (4 classes/labels). The work shows clearly that while using solely the non-intrusive EDA sensors a robust classification of human emotion is possible even without involving additional/other physiological signals

    A Domain-Independent Classification Model for Sentiment Analysis Using Neural Models

    No full text
    Most people nowadays depend on the Web as a primary source of information. Statistical studies show that young people obtain information mainly from Facebook, Twitter, and other social media platforms. By relying on these data, people may risk drawing the incorrect conclusions when reading the news or planning to buy a product. Therefore, systems that can detect and classify sentiments and assist users in finding the correct information on the Web is highly needed in order to prevent Web surfers from being easily deceived. This paper proposes an intensive study regarding domain-independent classification models for sentiment analysis that should be trained only once. The study consists of two phases: the first phase is based on a deep learning model which is training a neural network model once after extracting robust features and saving the model and its parameters. The second phase is based on applying the trained model on a totally new dataset, aiming at correctly classifying reviews as positive or negative. The proposed model is trained on the IMDb dataset and then tested on three different datasets: IMDb dataset, Movie Reviews dataset, and our own dataset collected from Amazon reviews that rate users’ opinions regarding Apple products. The work shows high performance using different evaluation metrics compared to the stat-of-the-art results

    HFM: A Hybrid Feature Model Based on Conditional Auto Encoders for Zero-Shot Learning

    Get PDF
    Zero-Shot Learning (ZSL) is related to training machine learning models capable of classifying or predicting classes (labels) that are not involved in the training set (unseen classes). A well-known problem in Deep Learning (DL) is the requirement for large amount of training data. Zero-Shot learning is a straightforward approach that can be applied to overcome this problem. We propose a Hybrid Feature Model (HFM) based on conditional autoencoders for training a classical machine learning model on pseudo training data generated by two conditional autoencoders (given the semantic space as a condition): (a) the first autoencoder is trained with the visual space concatenated with the semantic space and (b) the second autoencoder is trained with the visual space as an input. Then, the decoders of both autoencoders are fed by the test data of the unseen classes to generate pseudo training data. To classify the unseen classes, the pseudo training data are combined to train a support vector machine. Tests on four different benchmark datasets show that the proposed method shows promising results compared to the current state-of-the-art when it comes to settings for both standard Zero-Shot Learning (ZSL) and Generalized Zero-Shot Learning (GZSL)

    A Globally Generalized Emotion Recognition System Involving Different Physiological Signals

    No full text
    Machine learning approaches for human emotion recognition have recently demonstrated high performance. However, only/mostly for subject-dependent approaches, in a variety of applications like advanced driver assisted systems, smart homes and medical environments. Therefore, now the focus is shifted more towards subject-independent approaches, which are more universal and where the emotion recognition system is trained using a specific group of subjects and then tested on totally new persons and thereby possibly while using other sensors of same physiological signals in order to recognize their emotions. In this paper, we explore a novel robust subject-independent human emotion recognition system, which consists of two major models. The first one is an automatic feature calibration model and the second one is a classification model based on Cellular Neural Networks (CNN). The proposed system produces state-of-the-art results with an accuracy rate between 80% and 89% when using the same elicitation materials and physiological sensors brands for both training and testing and an accuracy rate of 71.05% when the elicitation materials and physiological sensors brands used in training are different from those used in training. Here, the following physiological signals are involved: ECG (Electrocardiogram), EDA (Electrodermal activity) and ST (Skin-Temperature)

    Zero-Shot Human Activity Recognition Using Non-Visual Sensors

    No full text
    Due to significant advances in sensor technology, studies towards activity recognition have gained interest and maturity in the last few years. Existing machine learning algorithms have demonstrated promising results by classifying activities whose instances have been already seen during training. Activity recognition methods based on real-life settings should cover a growing number of activities in various domains, whereby a significant part of instances will not be present in the training data set. However, to cover all possible activities in advance is a complex and expensive task. Concretely, we need a method that can extend the learning model to detect unseen activities without prior knowledge regarding sensor readings about those previously unseen activities. In this paper, we introduce an approach to leverage sensor data in discovering new unseen activities which were not present in the training set. We show that sensor readings can lead to promising results for zero-shot learning, whereby the necessary knowledge can be transferred from seen to unseen activities by using semantic similarity. The evaluation conducted on two data sets extracted from the well-known CASAS datasets show that the proposed zero-shot learning approach achieves a high performance in recognizing unseen (i.e., not present in the training dataset) new activities

    Modeling and Simulation of Shared Electric Automated and Connected Mobility Systems with Autonomous Repositioning: Performance Evaluation and Deployment

    No full text
    The boom seen in artificial intelligence in recent years has led to a revolution in the automotive industry. Numerous automakers around the world, such as Tesla, Toyota, Honda, and BMW, have achieved giant strides in the development of e-autonomous vehicles. Consequently, shared electric automated vehicle mobility (SEAVM) systems, which are a crucial part of future innovative transportation solutions, have attracted significant attention from the research community, particularly from a design perspective. However, the flexibility of shared automated mobility systems may lead to a self-operating technology issue (unequal distribution of vehicles), since users in these systems can pick up and drop off electric vehicles wherever they like. With this in mind, this paper addressed the issues of autonomous repositioning and the assignment of shared autonomous electric vehicle systems to balance a system’s network and fulfill its demand. Modeling, analysis and assessment of the system’s performance were carried out using stochastic Petri nets formalism, which included determining the average time areas were empty/congested and the number of unserved consumers, and estimating the redistribution service launch moment. Furthermore, many simulation scenarios were analyzed, including repositioning and without repositioning scenarios, in order to evaluate the efficiency of the model and to show the potential of using Petri nets as a probabilistic formalism approach for the modeling of e-automated mobility systems

    An Integral Projection-Based Semantic Autoencoder for Zero-Shot Learning

    No full text
    Zero-shot Learning (ZSL) classification categorizes or predicts classes (labels) that are not included in the training set (unseen classes). Recent works proposed different semantic autoencoder (SAE) models where the encoder embeds a visual feature vector space into the semantic space and the decoder reconstructs the original visual feature space. The objective is to learn the embedding by leveraging a source data distribution, which can be applied effectively to a different but related target data distribution. Such embedding-based methods are prone to domain shift problems and are vulnerable to biases. We propose an integral projection-based semantic autoencoder (IP-SAE) where an encoder projects a visual feature space concatenated with the semantic space into a latent representation space. We force the decoder to reconstruct the visual-semantic data space. Due to this constraint, the visual-semantic projection function preserves the discriminatory data included inside the original visual feature space. The enriched projection forces a more precise reconstitution of the visual feature space invariant to the domain manifold. Consequently, the learned projection function is less domain-specific and alleviates the domain shift problem. Our proposed IP-SAE model consolidates a symmetric transformation function for embedding and projection, and thus, it provides transparency for interpreting generative applications in ZSL. Therefore, in addition to outperforming state-of-the-art methods considering four benchmark datasets, our analytical approach allows us to investigate distinct characteristics of generative-based methods in the unique context of zero-shot inference

    Toward Authentication of Videos: Integer Transform Based Motion Vector Watermarking

    Get PDF
    Nowadays, digital content like videos, audio and images are widely used as evidence in criminal courts and forensic laboratories. Due to the advanced low-cost and easily available multimedia/communication tools and softwares, manipulation of the content is a no-brain task. Thus, the protection of digital content originality is a challenge for the content owners and researchers before it can be produced in court or used for some other purpose. We proposed a motion vector watermarking technique that validate and authenticate videos. We are embedding the correlated watermark in the integer wavelet transform domain. In our method, the selection of embedding areas is based on the variation of motion vectors. The video frames are fully protected in both spatial and transform domains since the watermark is correlated with the approximation subbands of wavelet transform before embedding. The proposed technique can concisely determine the attacked regions. The results validate the performance of the proposed approach in terms of quality metrics like peak signal-to-noise ratio (PSNR), structural similarity index (SSIM), normalized coefficients (NC) and bit error rate (BER)
    corecore